Explaining recommendations generated by MDPs

نویسندگان

  • Omar Zia Khan
  • Pascal Poupart
  • James P. Black
چکیده

There has been little work in explaining recommendations generated by Markov Decision Processes (MDPs). We analyze the difculty of explaining policies computed automatically and identify a set of templates that can be used to generate explanations automatically at run-time. These templates are domain-independent and can be used in any application of an MDP. We show that no additional e ort is required from the MDP designer for producing such explanations. We use the problem of advising undergraduate students in their course selection to explain the recommendation for selecting speci c courses to students. We also propose an extension to leverage domain-speci c constructs using ontologies so that explanations can be made more user-friendly.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Minimal Sufficient Explanations for Factored Markov Decision Processes

Explaining policies of Markov Decision Processes (MDPs) is complicated due to their probabilistic and sequential nature. We present a technique to explain policies for factored MDP by populating a set of domain-independent templates. We also present a mechanism to determine a minimal set of templates that, viewed together, completely justify the policy. Our explanations can be generated automat...

متن کامل

Policy Explanation and Model Refinement in Decision-Theoretic Planning

Decision-theoretic systems, such as Markov Decision Processes (MDPs), are used for sequential decision-making under uncertainty. MDPs provide a generic framework that can be applied in various domains to compute optimal policies. This thesis presents techniques that offer explanations of optimal policies for MDPs and then refine decision theoretic models (Bayesian networks and MDPs) based on fe...

متن کامل

Automated Explanations for MDP Policies

Explaining policies of Markov Decision Processes (MDPs) is complicated due to their probabilistic and sequential nature. We present a technique to explain policies for factored MDP by populating a set of domain-independent templates. We also present a mechanism to determine a minimal set of templates that, viewed together, completely justify the policy. We demonstrate our technique using the pr...

متن کامل

Abstraction and Approximate Decision-Theoretic Planning

ion and Approximate Decision Theoretic Planning Richard Dearden and Craig Boutiliery Department of Computer Science University of British Columbia Vancouver, British Columbia CANADA, V6T 1Z4 email: dearden,[email protected] Abstract Markov decision processes (MDPs) have recently been proposed as useful conceptual models for understanding decision-theoretic planning. However, the utility of the as...

متن کامل

An Empirical Analysis of Off-policy Learning in Discrete MDPs

Abstract Off-policy evaluation is the problem of evaluating a decision-making policy using data collected under a different behaviour policy. While several methods are available for addressing off-policy evaluation, little work has been done on identifying the best methods. In this paper, we conduct an in-depth comparative study of several off-policy evaluation methods in non-bandit, finite-hor...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008